# Subword embedding initialization
Roberta Base Wechsel German
MIT
A German RoBERTa model trained using the WECHSEL method, achieving cross-lingual transfer of monolingual language models through effective initialization of subword embeddings.
Large Language Model
Transformers German

R
benjamin
96
7
Gpt2 Wechsel French
MIT
A French version of GPT-2 trained using the WECHSEL method, achieving cross-lingual transfer of monolingual language models through effective initialization of subword embeddings.
Large Language Model
Transformers French

G
benjamin
33
4
Roberta Base Wechsel Swahili
MIT
A RoBERTa base model trained using the WECHSEL method, specifically optimized for Swahili to achieve efficient cross-lingual transfer.
Large Language Model
Transformers Other

R
benjamin
222
1
Roberta Base Wechsel Chinese
MIT
A RoBERTa Chinese model trained with the WECHSEL method, achieving efficient cross-lingual transfer from English to Chinese
Large Language Model
Transformers Chinese

R
benjamin
16
2
Gpt2 Wechsel Chinese
MIT
A Chinese GPT-2 model trained using the WECHSEL method, achieving cross-lingual transfer of monolingual language models through effective initialization of subword embeddings.
Large Language Model
Transformers Chinese

G
benjamin
19
4
Featured Recommended AI Models